# Multi-Precision Quantization
Deepseek R1 Llama 8B F32 GGUF
Apache-2.0
DeepSeek-R1-Llama-8B-F32-GGUF is the quantized version of DeepSeek-R1-Distill-Llama-8B, trained directly with reinforcement learning, featuring capabilities such as self-verification, reflection, and generating extended chain-of-thought reasoning.
Large Language Model
Transformers English

D
prithivMLmods
326
1
Microsoft Phi 4 Reasoning GGUF
MIT
This is a quantized version of Microsoft's Phi-4-reasoning model, optimized using llama.cpp for inference tasks and supporting multiple quantization options.
Large Language Model
M
bartowski
5,443
4
Google Gemma 3 12b It Qat GGUF
Gemma-3-12b model based on Google QAT (Quantization-Aware Training) weight quantization, offering multiple quantized versions to accommodate different hardware requirements.
Large Language Model
G
bartowski
10.78k
16
FLUX.1 Redux Dev GGUF
Other
FLUX.1-Redux-dev is a text-to-image generation model based on the FLUX technology stack, supporting the English language and adopting a non-commercial license.
Text-to-Image English
F
second-state
527
9
Bge Base En V1.5 Gguf
MIT
This project provides the BGE embedding model stored in GGUF format, which is suitable for use with llama.cpp and offers better performance than transformers.
Text Embedding
B
CompendiumLabs
1,108
5
Featured Recommended AI Models